BTLM-3B-8k-base is a 3-billion-parameter language model with an 8k context length, trained on the 627-billion-token SlimPajama dataset, delivering performance comparable to open-source 7-billion-parameter models.
Large Language Model
Transformers English